Equilibrium in misspecified Markov decision processes

نویسندگان

چکیده

We provide an equilibrium framework for modeling the behavior of agent who holds a simplified view dynamic optimization problem. The faces Markov decision process, where transition probability function determines evolution state variable as previous and agent's action. is uncertain about true has prior over set possible functions; this reflects (possibly simplified) her environment may not contain function. define concept conditions under which it characterizes steady?state when updates beliefs using Bayes' rule.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Equilibrium in Misspecified Markov Decision Processes ∗

We provide an equilibrium framework for modeling the behavior of an agent who holds a simplified view of a dynamic optimization problem. The agent faces a Markov Decision Process, where a transition probability function determines the evolution of a state variable as a function of the previous state and the agent’s action. The agent is uncertain about the true transition function and has a prio...

متن کامل

Equilibrium in Misspecified

We provide an equilibrium framework for modeling the behavior of an agent who holds a simplified view of a dynamic optimization problem. The agent faces a Markov Decision Process, where a transition probability function determines the evolution of a state variable as a function of the previous state and the agent’s action. The agent is uncertain about the true transition function and has a prio...

متن کامل

Bounded Parameter Markov Decision Processes Bounded Parameter Markov Decision Processes

In this paper, we introduce the notion of a bounded parameter Markov decision process as a generalization of the traditional exact MDP. A bounded parameter MDP is a set of exact MDPs speciied by giving upper and lower bounds on transition probabilities and rewards (all the MDPs in the set share the same state and action space). Bounded parameter MDPs can be used to represent variation or uncert...

متن کامل

Learning Qualitative Markov Decision Processes Learning Qualitative Markov Decision Processes

To navigate in natural environments, a robot must decide the best action to take according to its current situation and goal, a problem that can be represented as a Markov Decision Process (MDP). In general, it is assumed that a reasonable state representation and transition model can be provided by the user to the system. When dealing with complex domains, however, it is not always easy or pos...

متن کامل

Accelerated decomposition techniques for large discounted Markov decision processes

Many hierarchical techniques to solve large Markov decision processes (MDPs) are based on the partition of the state space into strongly connected components (SCCs) that can be classified into some levels. In each level, smaller problems named restricted MDPs are solved, and then these partial solutions are combined to obtain the global solution. In this paper, we first propose a novel algorith...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Theoretical Economics

سال: 2021

ISSN: ['1555-7561', '1933-6837']

DOI: https://doi.org/10.3982/te3843